SuperConText: Supervised Contrastive Learning Framework for Textual Representations

نویسندگان

چکیده

In the last decade, Deep neural networks (DNNs) have been proven to outperform conventional machine learning models in supervised tasks. Most of these are typically optimized by minimizing well-known Cross-Entropy objective function. The latter, however, has a number drawbacks, including poor margins and instability. Taking inspiration from recent self-supervised Contrastive representation approaches, we introduce Supervised framework for Textual representations (SuperConText) address those issues.We pretrain network novel fully-supervised contrastive loss. goal is increase both inter-class separability intra-class compactness embeddings latent space. Examples belonging same class regarded as positive pairs, while examples different classes considered negatives. Further, propose simple yet effective method selecting hard negatives during training phase. an extensive series experiments, study impact parameters on quality learned (e.g. batch size). Simulation results show that proposed solution outperforms several competing approaches various large-scale text classification benchmarks without requiring specialized architectures, data augmentations, memory banks, or additional unsupervised data. For instance, achieve top-1 accuracy 61.94% Amazon-F dataset, which 3.54% above best result obtained when using cross-entropy with model architecture.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Contrastive Learning of Emoji-based Representations for Resource-Poor Languages

The introduction of emojis (or emoticons) in social media platforms has given the users an increased potential for expression. We propose a novel method called Classification of Emojis using Siamese Network Architecture (CESNA) to learn emoji-based representations of resource-poor languages by jointly training them with resource-rich languages using a siamese network. CESNA model consists of tw...

متن کامل

DKPro TC: A Java-based Framework for Supervised Learning Experiments on Textual Data

We present DKPro TC, a framework for supervised learning experiments on textual data. The main goal of DKPro TC is to enable researchers to focus on the actual research task behind the learning problem and let the framework handle the rest. It enables rapid prototyping of experiments by relying on an easy-to-use workflow engine and standardized document preprocessing based on the Apache Unstruc...

متن کامل

Time-Contrastive Networks: Self-Supervised Learning from Video

We propose a self-supervised approach for learning representations and robotic behaviors entirely from unlabeled videos recorded from multiple viewpoints, and study how this representation can be used in two robotic imitation settings: imitating object interactions from videos of humans, and imitating human poses. Imitation of human behavior requires a viewpoint-invariant representation that ca...

متن کامل

Weakly-Supervised Learning with Cost-Augmented Contrastive Estimation

We generalize contrastive estimation in two ways that permit adding more knowledge to unsupervised learning. The first allows the modeler to specify not only the set of corrupted inputs for each observation, but also how bad each one is. The second allows specifying structural preferences on the latent variable used to explain the observations. They require setting additional hyperparameters, w...

متن کامل

Learning Semantic Textual Similarity with Structural Representations

Measuring semantic textual similarity (STS) is at the cornerstone of many NLP applications. Different from the majority of approaches, where a large number of pairwise similarity features are used to represent a text pair, our model features the following: (i) it directly encodes input texts into relational syntactic structures; (ii) relies on tree kernels to handle feature engineering automati...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2023

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2023.3241490